Skip to content

Conversation

@quic-indrbane
Copy link

…code

Reproduction steps:
transformers == 4.55.0

download alpaca dataset

run command
python -m QEfficient.cloud.finetune --device qaic --model_name
meta-llama/Llama-3.1-8B --tokenizer_name meta-llama/Llama-3.1-8B
--max_train_step 20 --output_dir
meta-llama/Llama-3.1-8B_alpaca_dataset_run --dataset alpaca_dataset
--run_validation False

Expected error:
TypeError: PreTrainedTokenizerFast._batch_encode_plus() got an unexpected keyword argument 'pad_to_max_length'

…code

Reproduction steps:
transformers == 4.55.0

download alpaca dataset

run command
python -m QEfficient.cloud.finetune --device qaic --model_name \
  meta-llama/Llama-3.1-8B --tokenizer_name meta-llama/Llama-3.1-8B \
  --max_train_step 20 --output_dir \
  meta-llama/Llama-3.1-8B_alpaca_dataset_run --dataset alpaca_dataset \
  --run_validation False

Expected error:
TypeError: PreTrainedTokenizerFast._batch_encode_plus() got an
unexpected keyword argument 'pad_to_max_length'

Signed-off-by: Indrajit Banerjee <indrbane@qti.qualcomm.com>
Copy link
Contributor

@quic-akuruvil quic-akuruvil left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hi @quic-indrbane 599 PR is already doing the same fix. Will merge that on priority to unblock the FT activities.

@quic-rishinr
Copy link
Contributor

can we close this PR as changes is added as part of #599 ?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants